Deberta Large Mnli Zero Cls
MIT
DeBERTa is an enhanced BERT decoding model based on the disentangled attention mechanism, surpassing BERT and RoBERTa in multiple natural language understanding tasks by improving the attention mechanism and masked decoder.
Large Language Model
Transformers English